521 research outputs found

    Locating 1-D Bar Codes in DCT-Domain

    Get PDF
    Today\u27s digital cameras and camera phones allow users to capture bar codes, which are used to uniquely identify consumer product. In this paper a fast algorithm is proposed that locates a 1-D bar code in the DCT-domain of a bar code image taken by a digital camera. The algorithm uses the DCT-transform properties to distinguish bar code from other texture, morphological operations to smooth the detected bar code area and the features of the extracted area lo detect position and orientation of a bar code in the imag

    Region segmentation for facial image compressing

    Get PDF
    This paper addresses the segmentation of passport images in order to improve quality of significant regions and to further reduce redundancy of insignificant ones. The approach is to first segment a facial image into two major regions, namely background and foreground. Here a new technique using pixel difference is presented. To compress facial regions at better quality, a face segmentation algorithm is introduced that detects eyes and mouth in a face. Region of interest (ROI) coding is then used to obtain better quality for facial features. At the end, some strategies that make use of region segmentation are proposed in order to increase performance in entropy codin

    JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

    Get PDF
    Two dimensional (2D) barcodes are becoming a pervasive interface for mobile devices, such as camera phones. Often, only monochrome 2D-barcodes are used due to their robustness in an uncontrolled operating environment of camera phones. Most camera phones capture and store such 2D-barcode images in the baseline JPEG format. As a lossy compression technique, JPEG does introduce a fair amount of error in the decoding of captured 2D-barcode images. In this paper, we introduce an improved JPEG compression scheme for such barcode images. By altering the JPEG compression parameters based on the DCT coefficient distribution of such barcode images, the improved compression scheme produces JPEG images with higher PSNR value as compared to the baseline implementation. We have also applied our improved scheme to a real 2D-barcode system - the QR Code and analyzed its performance against the baseline JPEG scheme

    Improving mobile color 2D-barcode JPEG image readability using DCT coefficient distributions

    Get PDF
    Two dimensional (2D) barcodes are becoming a pervasive interface for mobile devices, such as camera smartphones. Often, only monochrome 2D-barcodes are used due to their robustness in an uncontrolled operating environment of smartphones. Nonetheless, we are seeing an emerging use of color 2D-barcodes for camera smartphones. Most smartphones capture and store such 2D-barcode images in the baseline JPEG format. As a lossy compression technique, JPEG does introduce a fair amount of error in the captured 2D-barcode images. In this paper, we analyzed the Discrete Cosine Transform (DCT) coefficient distributions of generalized 2D-barcodes using colored data cells, each comprising of 4, 8 and 10 colors. Using these DCT distributions, we improved the JPEG compression of such mobile barcode images. By altering the JPEG compression parameters based on the DCT coefficient distribution of the barcode images, our improved compression scheme produces JPEG images with higher PSNR value as compared to the baseline implementation. We have also applied our improved scheme to a 10 colors 2D-barcode system; and analyzed its performance in comparison to the default and alternative JPEG schemes. We have found that our improved scheme does provide a marked improvement for the successful decoding of the 10 colors 2D-barcode system

    Texture-based latent space disentanglement for enhancement of a training dataset for ANN-based classification of fruit and vegetables

    Get PDF
    The capability of Convolutional Neural Networks (CNNs) for sparse representation has significant application to complex tasks like Representation Learning (RL). However, labelled datasets of sufficient size for learning this representation are not easily obtainable. The unsupervised learning capability of Variational Autoencoders (VAEs) and Generative Adversarial Networks (GANs) provide a promising solution to this issue through their capacity to learn representations for novel data samples and classification tasks. In this research, a texture-based latent space disentanglement technique is proposed to enhance learning of representations for novel data samples. A comparison is performed among different VAEs and GANs with the proposed approach for synthesis of new data samples. Two different VAE architectures are considered, a single layer dense VAE and a convolution based VAE, to compare the effectiveness of different architectures for learning of the representations. The GANs are selected based on the distance metric for disjoint distribution divergence estimation of complex representation learning tasks. The proposed texture-based disentanglement has been shown to provide a significant improvement for disentangling the process of representation learning by conditioning the random noise and synthesising texture rich images of fruit and vegetables

    A sample weight and adaboost CNN-based coarse to fine classification of fruit and vegetables at a supermarket self-checkout

    Get PDF
    © 2020 by the authors. Licensee MDPI, Basel, Switzerland. The physical features of fruit and vegetables make the task of vision-based classification of fruit and vegetables challenging. The classification of fruit and vegetables at a supermarket self-checkout poses even more challenges due to variable lighting conditions and human factors arising from customer interactions with the system along with the challenges associated with the colour, texture, shape, and size of a fruit or vegetable. Considering this complex application, we have proposed a progressive coarse to fine classification technique to classify fruit and vegetables at supermarket checkouts. The image and weight of fruit and vegetables have been obtained using a prototype designed to simulate the supermarket environment, including the lighting conditions. The weight information is used to change the coarse classification of 15 classes down to three, which are further used in AdaBoost-based Convolutional Neural Network (CNN) optimisation for fine classification. The training samples for each coarse class are weighted based on AdaBoost optimisation, which are updated on each iteration of a training phase. Multi-class likelihood distribution obtained by the fine classification stage is used to estimate a final classification with a softmax classifier. GoogleNet, MobileNet, and a custom CNN have been used for AdaBoost optimisation, with promising classification results

    Challenges and solutions for autonomous ground robot scene understanding and navigation in unstructured outdoor environments: A review

    Get PDF
    The capabilities of autonomous mobile robotic systems have been steadily improving due to recent advancements in computer science, engineering, and related disciplines such as cognitive science. In controlled environments, robots have achieved relatively high levels of autonomy. In more unstructured environments, however, the development of fully autonomous mobile robots remains challenging due to the complexity of understanding these environments. Many autonomous mobile robots use classical, learning-based or hybrid approaches for navigation. More recent learning-based methods may replace the complete navigation pipeline or selected stages of the classical approach. For effective deployment, autonomous robots must understand their external environments at a sophisticated level according to their intended applications. Therefore, in addition to robot perception, scene analysis and higher-level scene understanding (e.g., traversable/non-traversable, rough or smooth terrain, etc.) are required for autonomous robot navigation in unstructured outdoor environments. This paper provides a comprehensive review and critical analysis of these methods in the context of their applications to the problems of robot perception and scene understanding in unstructured environments and the related problems of localisation, environment mapping and path planning. State-of-the-art sensor fusion methods and multimodal scene understanding approaches are also discussed and evaluated within this context. The paper concludes with an in-depth discussion regarding the current state of the autonomous ground robot navigation challenge in unstructured outdoor environments and the most promising future research directions to overcome these challenges

    A comprehensive review of fruit and vegetable classification techniques

    Get PDF
    Recent advancements in computer vision have enabled wide-ranging applications in every field of life. One such application area is fresh produce classification, but the classification of fruit and vegetable has proven to be a complex problem and needs to be further developed. Fruit and vegetable classification presents significant challenges due to interclass similarities and irregular intraclass characteristics. Selection of appropriate data acquisition sensors and feature representation approach is also crucial due to the huge diversity of the field. Fruit and vegetable classification methods have been developed for quality assessment and robotic harvesting but the current state-of-the-art has been developed for limited classes and small datasets. The problem is of a multi-dimensional nature and offers significantly hyperdimensional features, which is one of the major challenges with current machine learning approaches. Substantial research has been conducted for the design and analysis of classifiers for hyperdimensional features which require significant computational power to optimise with such features. In recent years numerous machine learning techniques for example, Support Vector Machine (SVM), K-Nearest Neighbour (KNN), Decision Trees, Artificial Neural Networks (ANN) and Convolutional Neural Networks (CNN) have been exploited with many different feature description methods for fruit and vegetable classification in many real-life applications. This paper presents a critical comparison of different state-of-the-art computer vision methods proposed by researchers for classifying fruit and vegetable

    Review of deep learning methods in robotic grasp detection

    Get PDF
    For robots to attain more general-purpose utility, grasping is a necessary skill to master. Such general-purpose robots may use their perception abilities to visually identify grasps for a given object. A grasp describes how a robotic end-effector can be arranged to securely grab an object and successfully lift it without slippage. Traditionally, grasp detection requires expert human knowledge to analytically form the task-specific algorithm, but this is an arduous and time-consuming approach. During the last five years, deep learning methods have enabled significant advancements in robotic vision, natural language processing, and automated driving applications. The successful results of these methods have driven robotics researchers to explore the use of deep learning methods in task-generalised robotic applications. This paper reviews the current state-of-the-art in regards to the application of deep learning methods to generalised robotic grasping and discusses how each element of the deep learning approach has improved the overall performance of robotic grasp detection. Several of the most promising approaches are evaluated and the most suitable for real-time grasp detection is identified as the one-shot detection method. The availability of suitable volumes of appropriate training data is identified as a major obstacle for effective utilisation of the deep learning approaches, and the use of transfer learning techniques is proposed as a potential mechanism to address this. Finally, current trends in the field and future potential research directions are discussed

    Class distribution-aware adaptive margins and cluster embedding for classification of fruit and vegetables at supermarket self-checkouts

    Get PDF
    The complex task of vision based fruit and vegetables classification at a supermarket self-checkout poses significant challenges. These challenges include the highly variable physical features of fruit and vegetables i.e. colour, texture shape and size which are dependent upon ripeness and storage conditions in a supermarket as well as general product variation. Supermarket environments are also significantly variable with respect to lighting conditions. Attempting to build an exhaustive dataset to capture all these variations, for example a dataset of a fruit consisting of all possible colour variations, is nearly impossible. Moreover, some fruit and vegetable classes have significant similar physical features e.g. the colour and texture of cabbage and lettuce. Current state-of-the-art classification techniques such as those based on Deep Convolutional Neural Networks (DCNNs) are highly prone to errors resulting from the inter-class similarities and intra-class variations of fruit and vegetable images. The deep features of highly variable classes can invade the features of neighbouring similar classes in a learned feature space of the DCNN, resulting in confused classification hyper-planes. To overcome these limitations of current classification techniques we have proposed a class distribution-aware adaptive margins approach with cluster embedding for classification of fruit and vegetables. We have tested the proposed technique for cluster-based feature embedding and classification effectiveness. It is observed that introduction of adaptive classification margins proportional to the class distribution can achieve significant improvements in clustering and classification effectiveness. The proposed technique is tested for both clustering and classification, and promising results have been obtained
    • …
    corecore